Goto

Collaborating Authors

 Middleton


Multimodal Sleep Stage and Sleep Apnea Classification Using Vision Transformer: A Multitask Explainable Learning Approach

Kazemi, Kianoosh, Azimi, Iman, Khine, Michelle, Khayat, Rami N., Rahmani, Amir M., Liljeberg, Pasi

arXiv.org Artificial Intelligence

Sleep is an essential component of human physiology, contributing significantly to overall health and quality of life. Accurate sleep staging and disorder detection are crucial for assessing sleep quality. Studies in the literature have proposed PSG-based approaches and machine-learning methods utilizing single-modality signals. However, existing methods often lack multimodal, multilabel frameworks and address sleep stages and disorders classification separately. In this paper, we propose a 1D-Vision Transformer for simultaneous classification of sleep stages and sleep disorders. Our method exploits the sleep disorders' correlation with specific sleep stage patterns and performs a simultaneous identification of a sleep stage and sleep disorder. The model is trained and tested using multimodal-multilabel sensory data (including photoplethysmogram, respiratory flow, and respiratory effort signals). The proposed method shows an overall accuracy (cohen's Kappa) of 78% (0.66) for five-stage sleep classification and 74% (0.58) for sleep apnea classification. Moreover, we analyzed the encoder attention weights to clarify our models' predictions and investigate the influence different features have on the models' outputs. The result shows that identified patterns, such as respiratory troughs and peaks, make a higher contribution to the final classification process.


Fast Sampling generative model for Ultrasound image reconstruction

Lan, Hengrong, Li, Zhiqiang, He, Qiong, Luo, Jianwen

arXiv.org Artificial Intelligence

Image reconstruction from radio-frequency data is pivotal in ultrafast plane wave ultrasound imaging. Unlike the conventional delay-and-sum (DAS) technique, which relies on somewhat imprecise assumptions, deep learning-based methods perform image reconstruction by training on paired data, leading to a notable enhancement in image quality. Nevertheless, these strategies often exhibit limited generalization capabilities. Recently, denoising diffusion models have become the preferred paradigm for image reconstruction tasks. However, their reliance on an iterative sampling procedure results in prolonged generation time. In this paper, we propose a novel sampling framework that concurrently enforces data consistency of ultrasound signals and data-driven priors. By leveraging the advanced diffusion model, the generation of high-quality images is substantially expedited. Experimental evaluations on an in-vivo dataset indicate that our approach with a single plane wave surpasses DAS with spatial coherent compounding of 75 plane waves.


A Quantum Natural Language Processing Approach to Musical Intelligence

Miranda, Eduardo Reck, Yeung, Richie, Pearson, Anna, Meichanetzidis, Konstantinos, Coecke, Bob

arXiv.org Artificial Intelligence

There has been tremendous progress in Artificial Intelligence (AI) for music, in particular for musical composition and access to large databases for commercialisation through the Internet. We are interested in further advancing this field, focusing on composition. In contrast to current black-box AI methods, we are championing an interpretable compositional outlook on generative music systems. In particular, we are importing methods from the Distributional Compositional Categorical (DisCoCat) modelling framework for Natural Language Processing (NLP), motivated by musical grammars. Quantum computing is a nascent technology, which is very likely to impact the music industry in time to come. Thus, we are pioneering a Quantum Natural Language Processing (QNLP) approach to develop a new generation of intelligent musical systems. This work follows from previous experimental implementations of DisCoCat linguistic models on quantum hardware. In this chapter, we present Quanthoven, the first proof-of-concept ever built, which (a) demonstrates that it is possible to program a quantum computer to learn to classify music that conveys different meanings and (b) illustrates how such a capability might be leveraged to develop a system to compose meaningful pieces of music. After a discussion about our current understanding of music as a communication medium and its relationship to natural language, the chapter focuses on the techniques developed to (a) encode musical compositions as quantum circuits, and (b) design a quantum classifier. The chapter ends with demonstrations of compositions created with the system.


Virtual Agents in Live Coding: A Short Review

Xambó, Anna

arXiv.org Artificial Intelligence

AI and live coding has been little explored. This article contributes with a short review of different perspectives of using virtual agents in the practice of live coding looking at past and present as well as pointing to future directions.


SpCas9 activity prediction by DeepSpCas9, a deep learning–based model with high generalization performance

#artificialintelligence

To increase the accuracy of the analysis, deep sequencing data were filtered; target sequences with deep sequencing read counts below 200 and background indel frequencies above 8% were excluded as similarly performed previously (21). DNase-sequencing (DNase-seq) narrow peak data from ENCODE (36) were used to calculate chromatin accessibility as previously described (21). For each target site, 23 bases of the PAM plus protospacer sequence were aligned to the hg19 human reference genome using bowtie (41). Only the target sites that overlapped with DNase-seq narrow peaks were considered as DNase I hypersensitive target sites. We divided the Endo_Cas9 dataset into paired subsets by stratified random sampling from strata of DHS and non-DHS sites so that a similar ratio of DHS/non-DHS sites was assigned to each subset.


AI Tools That Help the Blind

WSJ.com: WSJD - Technology

Since losing his vision at age 13, Erik Weihenmayer has summited Mount Everest, white-water rafted and climbed frozen waterfalls. But making soup in his kitchen presented a unique challenge. On a frozen waterfall he could tap his ax against the ice to get a feel for its density, but in the kitchen, he had no way to differentiate between cans of tomato and chicken noodle. Mr. Weihenmayer, 49 years old, found a solution in Microsoft Corp.'s Seeing AI, a free app for the visually impaired. Among other things, the app can recognize faces, identify money, read handwriting and scan bar codes to differentiate between cans of soup.


Creative Partnerships with Technology: How Creativity Is Enhanced Through Interactions with Generative Computational Systems

Brown, Andrew Robert (Griffith University)

AAAI Conferences

This paper discusses emerging creative practices that involve interacting with generative computational systems, and the effect of such cybernetic interactions on our conceptions of creativity and agency. As computing systems have become more powerful in recent years, real time interaction with intelligent computational processes and models has emerged as a basis for innovative creative practices. Examples of these practices include interactive digital media installations, generative art works, live coding performances, virtual theatre, interactive cinema, and adaptive processes in computer games. In these types of activities computational systems have assumed a significant level of agency, or autonomy, that provoke questions about shared authorship and originality that are redefining our relationship with technologies and prompting new questions about human capabilities, values and the meaning of productive activities.